Adaptive Mixtures of Local Experts

نویسندگان

  • Robert A. Jacobs
  • Michael I. Jordan
  • Steven J. Nowlan
  • Geoffrey E. Hinton
چکیده

We present a new supervised learning procedure for systems composed of many separate networks, each of which learns to handle a subset of the complete set of training cases. The new procedure can be viewed either as a modular version of a multilayer supervised network, or as an associative version of competitive learning. It therefore provides a new link between these two apparently different approaches. We demonstrate that the learning procedure divides up a vowel discrimination task into appropriate subtasks, each of which can be solved by a very simple expert network. 1 Making associative learning competitive If backpropagation is used to train a single, multilayer network to perform different subtasks on different occasions, there will generally be strong interference effects which lead to slow learning and poor generalization. If we know in advance that a set of training cases may be naturally divided into subsets that correspond to distinct subtasks, interference can be reduced by using a system composed of several different “expert” networks plus a gating network that decides which of the experts should be used for each training case. 1 Hampshire This idea was first presented by Jacobs and Hinton at the Connectionist Summer School in Pittsburgh in 1988.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Predicting Flow Number of Asphalt Mixtures Based on the Marshall Mix design Parameters Using Multivariate Adaptive Regression Spline (MARS)

Rutting is one of the major distresses in the flexible pavements, which is heavily influenced by the asphalt mixtures properties at high temperatures. There are several methods for the characterization of the rutting resistance of asphalt mixtures. Flow number is one of the most important parameters that can be used for the evaluation of rutting. The flow number is measured by the dynamic creep...

متن کامل

Optimizing Mixtures of Local Experts in Tree-Like Regression Models

A mixture of local experts consists of a set of specialized models each of which is responsible for a particular local region of input space. Many algorithms in this class, for example M5 model tree, are sub-optimal (greedy). An algorithm for building optimal local mixtures of regression experts is proposed, and compared to MLP ANN on a number of cases.

متن کامل

The VC Dimension for Mixtures ofBinary Classi

The mixtures-of-experts (ME) methodology provides a tool of classii-cation when experts of logistic regression models or Bernoulli models are mixed according to a set of local weights. We show that the Vapnik-Chervonenkis (VC) dimension of the mixtures-of-experts architecture is bounded below by the number of experts m, and is bounded above by O(m 4 s 2), where s is the dimension of the input. ...

متن کامل

Adaptive neuro-fuzzy inference system (ANFIS) applied for spectrophotometric determination of fluoxetine and sertraline in pharmaceutical formulations and biological fluid

The UV-spectrophotometric method of analysis was proposed for simultaneous determination of fluoxetine (FLX) and sertraline (SRT). Considering the strong spectral overlap between UV-Vis spectra of these compounds, a previous separation should be carried out in order to determine them by conventional spectrophotometric techniques. Here, full-spectrum multivariate calibrations adaptive neuro-fuzz...

متن کامل

On the asymptotic normality of hierarchical mixtures-of-experts for generalized linear models

| In the class of hierarchical mixtures-of-experts (HME) models, \experts" in the exponential family with generalized linear mean functions of the form (+ x T) are mixed, according to a set of local weights called the \gating functions" depending on the predictor x. Here () is the inverse link function. We provide regularity conditions on the experts and on the gating functions under which the ...

متن کامل

On the identifiability of mixtures-of-experts

In mixtures-of-experts (ME) models, "experts" of generalized linear models are combined, according to a set of local weights called the "gating function". The invariant transformations of the ME probability density functions include the permutations of the expert labels and the translations of the parameters in the gating functions. Under certain conditions, we show that the ME systems are iden...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Neural Computation

دوره 3  شماره 

صفحات  -

تاریخ انتشار 1991